144 research outputs found

    EMPATH: A Neural Network that Categorizes Facial Expressions

    Get PDF
    There are two competing theories of facial expression recognition. Some researchers have suggested that it is an example of "categorical perception." In this view, expression categories are considered to be discrete entities with sharp boundaries, and discrimination of nearby pairs of expressive faces is enhanced near those boundaries. Other researchers, however, suggest that facial expression perception is more graded and that facial expressions are best thought of as points in a continuous, low-dimensional space, where, for instance, "surprise" expressions lie between "happiness" and "fear" expressions due to their perceptual similarity. In this article, we show that a simple yet biologically plausible neural network model, trained to classify facial expressions into six basic emotions, predicts data used to support both of these theories. Without any parameter tuning, the model matches a variety of psychological data on categorization, similarity, reaction times, discrimination, and recognition difficulty, both qualitatively and quantitatively. We thus explain many of the seemingly complex psychological phenomena related to facial expression perception as natural consequences of the tasks' implementations in the brain

    Bayesian-Augmented Identification of Stars in a Narrow View

    Get PDF
    An algorithm for the identification of stars from a charge-coupled-device (CCD) image of a star field has been extended for use with narrower field-of-view images. Previously, the algorithm had been shown to be effective at a field of view of 8 degrees. This work augments the earlier algorithm using Bayesian decision theory. The new algorithm is shown to be capable of effective star identification down to a field of view of 2 degrees. The algorithm was developed for use in estimating the attitude of a spacecraft and could be used on Earth to help in the identification of stars and other celestial objects for astronomical observations

    Holographic aberration correction: optimising the stiffness of an optical trap deep in the sample

    Get PDF
    We investigate the effects of 1st order spherical aberration and defocus upon the stiffness of an optical trap tens of μm into the sample. We control both these aberrations with a spatial light modulator. The key to maintain optimum trap stiffness over a range of depths is a specific non-trivial combination of defocus and axial objective position. This optimisation increases the trap stiffness by up to a factor of 3 and allows trapping of 1μm polystyrene beads up to 50μm deep in the sample.<p></p&gt

    Mapping Nearby Terrain in 3D by Use of a Grid of Laser Spots

    Get PDF
    A proposed optoelectronic system, to be mounted aboard an exploratory robotic vehicle, would be used to generate a three-dimensional (3D) map of nearby terrain and obstacles for purposes of navigating the vehicle across the terrain and avoiding the obstacles. The difference between this system and the other systems would lie in the details of implementation. In this system, the illumination would be provided by a laser. The beam from the laser would pass through a two-dimensional diffraction grating, which would divide the beam into multiple beams propagating in different, fixed, known directions. These beams would form a grid of bright spots on the nearby terrain and obstacles. The centroid of each bright spot in the image would be computed. For each such spot, the combination of (1) the centroid, (2) the known direction of the light beam that produced the spot, and (3) the known baseline would constitute sufficient information for calculating the 3D position of the spot

    Automated Camera Array Fine Calibration

    Get PDF
    Using aerial imagery, the JPL FineCalibration (JPL FineCal) software automatically tunes a set of existing CAHVOR camera models for an array of cameras. The software finds matching features in the overlap region between images from adjacent cameras, and uses these features to refine the camera models. It is not necessary to take special imagery of a known target and no surveying is required. JPL FineCal was developed for use with an aerial, persistent surveillance platform

    High-resolution, continuous field-of-view (FOV), non-rotating imaging system

    Get PDF
    A high resolution CMOS imaging system especially suitable for use in a periscope head. The imaging system includes a sensor head for scene acquisition, and a control apparatus inclusive of distributed processors and software for device-control, data handling, and display. The sensor head encloses a combination of wide field-of-view CMOS imagers and narrow field-of-view CMOS imagers. Each bank of imagers is controlled by a dedicated processing module in order to handle information flow and image analysis of the outputs of the camera system. The imaging system also includes automated or manually controlled display system and software for providing an interactive graphical user interface (GUI) that displays a full 360-degree field of view and allows the user or automated ATR system to select regions for higher resolution inspection

    Constructing a Database from Multiple 2D Images for Camera Pose Estimation and Robot Localization

    Get PDF
    The LMDB (Landmark Database) Builder software identifies persistent image features (landmarks) in a scene viewed multiple times and precisely estimates the landmarks 3D world positions. The software receives as input multiple 2D images of approximately the same scene, along with an initial guess of the camera poses for each image, and a table of features matched pair-wise in each frame. LMDB Builder aggregates landmarks across an arbitrarily large collection of frames with matched features. Range data from stereo vision processing can also be passed to improve the initial guess of the 3D point estimates. The LMDB Builder aggregates feature lists across all frames, manages the process to promote selected features to landmarks, and iteratively calculates the 3D landmark positions using the current camera pose estimations (via an optimal ray projection method), and then improves the camera pose estimates using the 3D landmark positions. Finally, it extracts image patches for each landmark from auto-selected key frames and constructs the landmark database. The landmark database can then be used to estimate future camera poses (and therefore localize a robotic vehicle that may be carrying the cameras) by matching current imagery to landmark database image patches and using the known 3D landmark positions to estimate the current pose

    A Multi-Stage Neural Network for Automatic Target Detection

    Get PDF
    ©1998 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other users, including reprinting/ republishing this material for advertising or promotional purposes, creating new collective works for resale or redistribution to servers or lists, or reuse of any copyrighted components of this work in other works.Presented at the 1998 International Joint Conference on Neural Networks (IJCNN) Anchorage, Alaska, May 1998.DOI: 10.1109/IJCNN.1998.682268Automatic Target Recognition (ATR) involves processing two-dimensional images for detecting, classifying, and tracking targets. The first stage in ATR is the detection process. This involves discrimination between target and non-target objects in a scene. In this paper, we shall discuss a novel approach which addresses the target detection process. This method extracts relevant object features utilizing principal component analysis. These extracted features are then presented to a multi-stage neural network which allows an overall increase in detection rate, while decreasing the false positive alarm rate. We shall discuss the techniques involved and present some detection results that have been implemented on the multi-stage neural network

    Real-Time Feature Tracking Using Homography

    Get PDF
    This software finds feature point correspondences in sequences of images. It is designed for feature matching in aerial imagery. Feature matching is a fundamental step in a number of important image processing operations: calibrating the cameras in a camera array, stabilizing images in aerial movies, geo-registration of images, and generating high-fidelity surface maps from aerial movies. The method uses a Shi-Tomasi corner detector and normalized cross-correlation. This process is likely to result in the production of some mismatches. The feature set is cleaned up using the assumption that there is a large planar patch visible in both images. At high altitude, this assumption is often reasonable. A mathematical transformation, called an homography, is developed that allows us to predict the position in image 2 of any point on the plane in image 1. Any feature pair that is inconsistent with the homography is thrown out. The output of the process is a set of feature pairs, and the homography. The algorithms in this innovation are well known, but the new implementation improves the process in several ways. It runs in real-time at 2 Hz on 64-megapixel imagery. The new Shi-Tomasi corner detector tries to produce the requested number of features by automatically adjusting the minimum distance between found features. The homography-finding code now uses an implementation of the RANSAC algorithm that adjusts the number of iterations automatically to achieve a pre-set probability of missing a set of inliers. The new interface allows the caller to pass in a set of predetermined points in one of the images. This allows the ability to track the same set of points through multiple frames
    • …
    corecore